"main.py --dataset Cora --hidden_dim 256 --epoch 30 --lr 0.01 --runtime 1 --classifier_time 10 --classifier_epochs 2000 --cuda 1 --classifier_lr 0.015 --classifier_decay 0.001 --batch_norm --loss cossum --a1 0.5 --a2 0.5 --classifier_seed 150 --GCN_layers_1 1"
test_acc: 0.7726
test_std: 0.001356465996625055
test_acc_mean: 0.7726
test_acc_std: 0.0
{
  "classifier_epochs": 2000,
  "batch_size": 1,
  "patience": 200,
  "l2_coef": 0.0,
  "drop_prob": 0.0,
  "dataset": "Cora",
  "num_hop": 0,
  "cuda": 1,
  "seed": 1234,
  "classifier_lr": 0.015,
  "classifier_decay": 0.001,
  "classifier_seed": 150,
  "classifier_time": 10,
  "classifier": null,
  "hidden_dim": 256,
  "proj_dim": 256,
  "fc_out": false,
  "input_dropout": 0.4,
  "dropout": 0.4,
  "batch_norm": true,
  "optimizer": "adam",
  "lr": 0.01,
  "decay": 0.0005,
  "epoch": 30,
  "alpha_dim": "sc",
  "no_alpha_sigmoid": false,
  "beta_dim": "sc",
  "block": "constant",
  "function": "laplacian",
  "use_mlp": false,
  "add_source": false,
  "time": 3.0,
  "augment": false,
  "method": "euler",
  "step_size": 1.0,
  "max_iters": 100,
  "adjoint_method": "adaptive_heun",
  "adjoint": false,
  "adjoint_step_size": 1,
  "tol_scale": 1.0,
  "tol_scale_adjoint": 1.0,
  "ode_blocks": 1,
  "max_nfe": 1000,
  "no_early": false,
  "earlystopxT": 3,
  "max_test_steps": 100,
  "beltrami": false,
  "leaky_relu_slope": 0.2,
  "attention_dropout": 0.0,
  "heads": 4,
  "attention_norm_idx": 0,
  "attention_dim": 16,
  "mix_features": false,
  "reweight_attention": false,
  "attention_type": "scaled_dot",
  "square_plus": false,
  "data_norm": "gcn",
  "self_loop_weight": 1.0,
  "runtime": 1,
  "loss": "cossum",
  "random_splits": false,
  "alpha_ode": 1.0,
  "alpha_ode_1": 1.0,
  "alpha_ode_2": 0.1,
  "a1": 0.5,
  "a2": 0.5,
  "reaction_term": "bspm",
  "beta_diag": false,
  "lambda": 0.1,
  "GCN_layers_1": 1,
  "GCN_layers_2": 2,
  "GAT_layers_1": 2,
  "GAT_layers_2": 2,
  "num_head": 2,
  "num_nodes": 2708
}
